9 research outputs found
Cross Lingual Transfer Learning for Zero-Resource Domain Adaptation
We propose a method for zero-resource domain adaptation of DNN acoustic
models, for use in low-resource situations where the only in-language training
data available may be poorly matched to the intended target domain. Our method
uses a multi-lingual model in which several DNN layers are shared between
languages. This architecture enables domain adaptation transforms learned for
one well-resourced language to be applied to an entirely different low-resource
language. First, to develop the technique we use English as a well-resourced
language and take Spanish to mimic a low-resource language. Experiments in
domain adaptation between the conversational telephone speech (CTS) domain and
broadcast news (BN) domain demonstrate a 29% relative WER improvement on
Spanish BN test data by using only English adaptation data. Second, we
demonstrate the effectiveness of the method for low-resource languages with a
poor match to the well-resourced language. Even in this scenario, the proposed
method achieves relative WER improvements of 18-27% by using solely English
data for domain adaptation. Compared to other related approaches based on
multi-task and multi-condition training, the proposed method is able to better
exploit well-resource language data for improved acoustic modelling of the
low-resource target domain.Comment: Submitted to ICASSP 2020. Main updates wrt previous versions: same
network config in all experiments, added Babel/Material LR target language
experiments, added comparison with alternative/similar methods of
cross-lingual adaptatio
The Edinburgh International Accents of English Corpus: Towards the Democratization of English ASR
English is the most widely spoken language in the world, used daily by
millions of people as a first or second language in many different contexts. As
a result, there are many varieties of English. Although the great many advances
in English automatic speech recognition (ASR) over the past decades, results
are usually reported based on test datasets which fail to represent the
diversity of English as spoken today around the globe. We present the first
release of The Edinburgh International Accents of English Corpus (EdAcc). This
dataset attempts to better represent the wide diversity of English,
encompassing almost 40 hours of dyadic video call conversations between
friends. Unlike other datasets, EdAcc includes a wide range of first and
second-language varieties of English and a linguistic background profile of
each speaker. Results on latest public, and commercial models show that EdAcc
highlights shortcomings of current English ASR models. The best performing
model, trained on 680 thousand hours of transcribed data, obtains an average of
19.7% word error rate (WER) -- in contrast to the 2.7% WER obtained when
evaluated on US English clean read speech. Across all models, we observe a drop
in performance on Indian, Jamaican, and Nigerian English speakers. Recordings,
linguistic backgrounds, data statement, and evaluation scripts are released on
our website (https://groups.inf.ed.ac.uk/edacc/) under CC-BY-SA license.Comment: Accepted to IEEE ICASSP 202